Zero-Variance Importance Sampling Estimators for Markov Process Expectations
نویسندگان
چکیده
We study the structure of zero-variance importance sampling estimators for expectations of functionals of Markov processes. For a class of expectations that can be characterized as solutions to linear systems, we show that a zerovariance estimator can be constructed by using an importance distribution that preserves the Markovian nature of the underlying process. This suggests that good practical importance sampling distributions can be found by searching within the class of Markovian probability distributions. The class of expectations considered includes as particular cases, among others: the mean time until hitting a rare set, the expected cumulative discounted reward until hitting a set, the mean duration of an excursion, the transient (deterministic horizon) expectation of a discounted final payoff, and steady-state expectations.
منابع مشابه
Variance reduction of estimators arising from Metropolis-Hastings algorithms
The Metropolis–Hastings algorithm is one of the most basic and well-studied Markov chain Monte Carlo methods. It generates a Markov chain which has as limit distribution the target distribution by simulating observations from a different proposal distribution. A proposed value is accepted with some particular probability otherwise the previous value is repeated. As a consequence, the accepted v...
متن کاملImportance Sampling in Path Space for Diffusion Processes
Importance sampling is a widely used technique to reduce the variance of the Monte Carlo method. It uses the idea of change of measure to design efficient Monte Carlo estimators. In this work, we study the importance sampling method in the framework of diffusion process and consider the change of measures which can be realized by adding a control force to the original dynamics. For certain expo...
متن کاملEstimating standard errors for importance sampling estimators with multiple Markov chains
The naive importance sampling estimator based on the samples from a single importance density can be extremely numerically unstable. We consider multiple distributions importance sampling estimators where samples from more than one probability distributions are combined to consistently estimate means with respect to given target distributions. These generalized importance sampling estimators pr...
متن کامل7 Importance Tempering
Simulated tempering (ST) is an established Markov Chain Monte Carlo (MCMC) methodology for sampling from a multimodal density π(θ). The technique involves introducing an auxiliary variable k taking values in a finite subset of [0, 1] and indexing a set of tempered distributions, say π k (θ) ∝ π(θ) k. Small values of k encourage better mixing, but samples from π are only obtained when the joint ...
متن کاملA ug 2 00 7 Importance Tempering
Simulated tempering (ST) is an established Markov Chain Monte Carlo (MCMC) methodology for sampling from a multimodal density π(θ). The technique involves introducing an auxiliary variable k taking values in a finite subset of [0, 1] and indexing a set of tempered distributions, say π k (θ) ∝ π(θ) k. Small values of k encourage better mixing, but samples from π are only obtained when the joint ...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
- Math. Oper. Res.
دوره 38 شماره
صفحات -
تاریخ انتشار 2013